# Efficient Text Generation

Simplestories 30M
MIT
SimpleStories is a micro model family designed for interpretability research, trained on the SimpleStories dataset, focusing on story generation tasks.
Text Generation Safetensors English
S
SimpleStories
735
1
Huihui Ai.glm 4 9B 0414 Abliterated GGUF
GLM-4-9B-0414-abliterated is a large language model with 9B parameters based on the GLM architecture, suitable for text generation tasks.
Large Language Model
H
DevQuasar
3,172
3
Orpheus 3b 0.1 Ft Q4 K M GGUF
Apache-2.0
This model is a GGUF-format conversion of canopylabs/orpheus-3b-0.1-ft, suitable for text generation tasks.
Large Language Model English
O
athenasaurav
162
0
Gemma 3 27b It GGUF
GGUF quantized version of the Gemma 3 27B instruction-tuned model, suitable for text generation tasks.
Large Language Model
G
MaziyarPanahi
475.91k
4
Huihui Ai.deepseek R1 Distill Qwen 32B Abliterated GGUF
This is a quantized version of a large language model, aiming to make knowledge more accessible and usable.
Large Language Model
H
DevQuasar
572
3
Deepseek Ai.deepseek R1 Distill Llama 8B GGUF
DeepSeek-R1-Distill-Llama-8B is an 8B-parameter large language model based on the Llama architecture, optimized through distillation training for text generation tasks.
Large Language Model
D
DevQuasar
320
3
Miniplm Qwen 200M
Apache-2.0
A 200M-parameter model based on the Qwen architecture, pretrained from scratch using the MiniPLM knowledge distillation framework
Large Language Model Transformers English
M
MiniLLM
203
5
Hymba 1.5B Base
Other
Hymba-1.5B-Base is a foundational text generation model developed by NVIDIA, employing a hybrid architecture combining Mamba and attention heads, suitable for various natural language generation tasks.
Large Language Model Transformers
H
nvidia
3,492
142
Stick OCR V4
MIT
This is a multilingual image-to-text model customized for internal use cases, supporting both English and German.
Large Language Model Transformers Supports Multiple Languages
S
DaMax96
597
1
Powerlm 3b
Apache-2.0
PowerLM-3B is a small language model with 3 billion parameters, trained using the Power learning rate scheduler, and demonstrates excellent performance across multiple benchmarks including natural language multiple-choice, code generation, and mathematical reasoning.
Large Language Model Transformers
P
ibm-research
11.07k
20
Layerskip Llama2 7B
Other
An improved model based on Llama2 7B, supporting hierarchical skip and self-speculative decoding to enhance inference efficiency
Large Language Model Transformers English
L
facebook
1,674
14
Mistral 7B Instruct V0.3 GGUF
Apache-2.0
GGUF quantized version of Mistral-7B-Instruct-v0.3, suitable for local inference text generation models
Large Language Model
M
MaziyarPanahi
253.99k
93
Mistral 7B V0.3
Apache-2.0
Mistral-7B-v0.3 is an upgraded large language model based on Mistral-7B-v0.2, with the main improvement being the expansion of the vocabulary to 32,768 tokens.
Large Language Model Transformers
M
mistralai
442.55k
472
Mistral 7b V0.3 Summarizer
Mistral-7B-Instruct-v0.3 is an instruction-tuned version of Mistral-7B, focusing on text generation tasks that follow human instructions.
Large Language Model Transformers English
M
devesh-2002
22
0
Mamba 7b Rw
Apache-2.0
Mamba-7B is a 7-billion-parameter model based on the Mamba architecture, trained over multiple rounds on the RefinedWeb dataset (1.2 trillion tokens). Mamba is a state space model that does not use self-attention mechanisms and excels in various natural language benchmarks.
Large Language Model English
M
TRI-ML
188
55
Mamba 130m Hf
Mamba is a transformer-compatible sequence modeling model with efficient inference capabilities.
Large Language Model Transformers
M
state-spaces
46.83k
56
Mamba 1.4b Hf
Mamba is an efficient language model based on the State Space Model (SSM) architecture, with 1.4B parameters, supporting text generation tasks
Large Language Model Transformers
M
state-spaces
5,431
11
Darebeagle 7B
Apache-2.0
DareBeagle-7B is a 7B-parameter large language model obtained by merging mlabonne/NeuralBeagle14-7B and mlabonne/NeuralDaredevil-7B using LazyMergekit, demonstrating excellent performance across multiple benchmarks.
Large Language Model Transformers
D
shadowml
77
1
Decilm 7B
Apache-2.0
DeciLM-7B is a pure decoder-based text generation model with 7.04 billion parameters, released under the Apache 2.0 license. It is currently the best-performing 7B base language model on the Open LLM Leaderboard.
Large Language Model Transformers English
D
Deci
2,833
226
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase